Rate of convergence of penalized likelihood context tree estimators
نویسنده
چکیده
The Bayesian Information Criterion (BIC) was first proposed by Schwarz (1978) as a model selection technique. It was thought that BIC was not appropriate for the case of context tree estimation, because of the huge number of trees that has to be tested. Recently, Csiszár & Talata (2006) proved the almost surely consistency of the BIC estimator and they also showed that it can be computed in linear time. Nevertheless, the rate of convergence of the BIC estimator remained as an open question, even in the simpler form for estimating the order of a Markov chain. The latter is a particular case of the tree estimation problem and can be derived from our results. The paper is organized as follows. In Section 2 we introduce some definitions and state the main results. In Section 3 we obtain exponential inequalities for empirical counts and empirical transition probabilities, generalizing a previous result in Galves et al. (2006) to the case of unbounded trees. These results are the key in the proof of the rate of convergence of the BIC estimator and by their relevant importance we included them in a separate section. Finally, in Section 4 we proved the main results in this paper and in Section 5 we present some discussion.
منابع مشابه
On the Rate of Convergence of Penalized Likelihood Context Tree Estimators
Abstract. We find an upper bound for the probability of error of penalized likelihood context tree estimators. The bound is explicit and applies to processes of unbounded memory, that constitute a subclass of infinite memory processes. We show that the maximal exponential decay for the probability of error is typically achieved with a penalizing term of the form n/log(n), where n is the sample ...
متن کاملSome upper bounds for the rate of convergence of penalized likelihood context tree estimators
Abstract. We find upper bounds for the probability of underestimation and overestimation errors in penalized likelihood context tree estimation. The bounds are explicit and applies to processes of not necessarily finite memory. We allow for general penalizing terms and we give conditions over the maximal depth of the estimated trees in order to get strongly consistent estimates. This generalize...
متن کاملOracle Inequalities and Adaptive Rates
We have previously seen how sieve estimators give rise to rates of convergence to the Bayes risk by performing empirical risk minimization over Hk(n), where (Hk)k ≥ 1 is an increasing sequence of sets of classifiers, and k(n) → ∞. However, the rate of convergence depends on k(n). Usually this rate is chosen to minimize the worst-case rate over all distributions of interest. However, it would be...
متن کاملPenalized Likelihood-type Estimators for Generalized Nonparametric Regression
We consider the asymptotic analysis of penalized likelihood type estimators for generalized non-parametric regression problems in which the target parameter is a vector valued function defined in terms of the conditional distribution of a response given a set of covariates. A variety of examples including ones related to generalized linear models and robust smoothing are covered by the theory. ...
متن کاملGeneralized Nonparametric Regression via Penalized Likelihood
We consider the asymptotic analysis of penalized likelihood type estimators for generalized non-parametric regression problems in which the target parameter is a vector valued function defined in terms of the conditional distribution of a response given a set of covariates, A variety of examples including ones related to generalized linear models and robust smoothing are covered by the theory. ...
متن کامل